58 research outputs found

    Beyond Fairness: Age-Harmless Parkinson's Detection via Voice

    Full text link
    Parkinson's disease (PD), a neurodegenerative disorder, often manifests as speech and voice dysfunction. While utilizing voice data for PD detection has great potential in clinical applications, the widely used deep learning models currently have fairness issues regarding different ages of onset. These deep models perform well for the elderly group (age >> 55) but are less accurate for the young group (age \leq 55). Through our investigation, the discrepancy between the elderly and the young arises due to 1) an imbalanced dataset and 2) the milder symptoms often seen in early-onset patients. However, traditional debiasing methods are impractical as they typically impair the prediction accuracy for the majority group while minimizing the discrepancy. To address this issue, we present a new debiasing method using GradCAM-based feature masking combined with ensemble models, ensuring that neither fairness nor accuracy is compromised. Specifically, the GradCAM-based feature masking selectively obscures age-related features in the input voice data while preserving essential information for PD detection. The ensemble models further improve the prediction accuracy for the minority (young group). Our approach effectively improves detection accuracy for early-onset patients without sacrificing performance for the elderly group. Additionally, we propose a two-step detection strategy for the young group, offering a practical risk assessment for potential early-onset PD patients

    Family functioning as a moderator in the relation between perceived stress and psychotic-like experiences among adolescents during COVID-19

    Get PDF
    Background: The COVID-19 pandemic has increased psychological stress among adolescents, and the relation between perceived stress (PS) and psychotic-like experiences (PLEs) has been well-established. However, little is known about the role of family functioning (FF) in this relation, especially when adolescents experienced the extended lockdown period with family members. Methods: A total of 4807 adolescents completed this retrospective paper-and-pencil survey after school reopening between May 14th and June 6th, 2020 in Hunan Province, China. We measured PS with the Perceived stress scale (PSS-10), PLEs with the eight positive items from Community Assessment of Psychic Experiences (CAPE-8), and FF with the Family APGAR scale. We conducted subgroup analysis based on three FF levels (good, moderate, and poor) determined by previous studies. Finally, correlation and moderation analysis were performed to detect the effect of FF in the relation between PS and PLEs after adjusting for demographic variables. Results: Adolescents with poor FF had higher levels of PS and higher prevalence of PLEs compared to those with good FF (both p \u3c 0.001). FF was negatively associated with both PS (r = −0.34, p \u3c 0.001) and PLEs (r = −0.29, p \u3c 0.001). Higher FF significantly attenuated the effect of PS on PLEs after adjusting for sex and age (effect = −0.011, bootstrap 95% CI -0.018, −0.005). Conclusion: Our findings indicate that well-functioned family could protect against stress-induced PLEs among adolescents during this crisis. Thus family system could be an early interventional target for distressing psychotic-like experiences in youngsters

    MINER: Improving Out-of-Vocabulary Named Entity Recognition from an Information Theoretic Perspective

    Full text link
    NER model has achieved promising performance on standard NER benchmarks. However, recent studies show that previous approaches may over-rely on entity mention information, resulting in poor performance on out-of-vocabulary (OOV) entity recognition. In this work, we propose MINER, a novel NER learning framework, to remedy this issue from an information-theoretic perspective. The proposed approach contains two mutual information-based training objectives: i) generalizing information maximization, which enhances representation via deep understanding of context and entity surface forms; ii) superfluous information minimization, which discourages representation from rote memorizing entity names or exploiting biased cues in data. Experiments on various settings and datasets demonstrate that it achieves better performance in predicting OOV entities

    Unsupervised Summarization for Chat Logs with Topic-Oriented Ranking and Context-Aware Auto-Encoders

    Full text link
    Automatic chat summarization can help people quickly grasp important information from numerous chat messages. Unlike conventional documents, chat logs usually have fragmented and evolving topics. In addition, these logs contain a quantity of elliptical and interrogative sentences, which make the chat summarization highly context dependent. In this work, we propose a novel unsupervised framework called RankAE to perform chat summarization without employing manually labeled data. RankAE consists of a topic-oriented ranking strategy that selects topic utterances according to centrality and diversity simultaneously, as well as a denoising auto-encoder that is carefully designed to generate succinct but context-informative summaries based on the selected utterances. To evaluate the proposed method, we collect a large-scale dataset of chat logs from a customer service environment and build an annotated set only for model evaluation. Experimental results show that RankAE significantly outperforms other unsupervised methods and is able to generate high-quality summaries in terms of relevance and topic coverage.Comment: Accepted by AAAI 2021, 9 page

    Topic-Oriented Spoken Dialogue Summarization for Customer Service with Saliency-Aware Topic Modeling

    Full text link
    In a customer service system, dialogue summarization can boost service efficiency by automatically creating summaries for long spoken dialogues in which customers and agents try to address issues about specific topics. In this work, we focus on topic-oriented dialogue summarization, which generates highly abstractive summaries that preserve the main ideas from dialogues. In spoken dialogues, abundant dialogue noise and common semantics could obscure the underlying informative content, making the general topic modeling approaches difficult to apply. In addition, for customer service, role-specific information matters and is an indispensable part of a summary. To effectively perform topic modeling on dialogues and capture multi-role information, in this work we propose a novel topic-augmented two-stage dialogue summarizer (TDS) jointly with a saliency-aware neural topic model (SATM) for topic-oriented summarization of customer service dialogues. Comprehensive studies on a real-world Chinese customer service dataset demonstrated the superiority of our method against several strong baselines.Comment: Accepted by AAAI 2021, 9 page

    TRACE: A Comprehensive Benchmark for Continual Learning in Large Language Models

    Full text link
    Aligned large language models (LLMs) demonstrate exceptional capabilities in task-solving, following instructions, and ensuring safety. However, the continual learning aspect of these aligned LLMs has been largely overlooked. Existing continual learning benchmarks lack sufficient challenge for leading aligned LLMs, owing to both their simplicity and the models' potential exposure during instruction tuning. In this paper, we introduce TRACE, a novel benchmark designed to evaluate continual learning in LLMs. TRACE consists of 8 distinct datasets spanning challenging tasks including domain-specific tasks, multilingual capabilities, code generation, and mathematical reasoning. All datasets are standardized into a unified format, allowing for effortless automatic evaluation of LLMs. Our experiments show that after training on TRACE, aligned LLMs exhibit significant declines in both general ability and instruction-following capabilities. For example, the accuracy of llama2-chat 13B on gsm8k dataset declined precipitously from 28.8\% to 2\% after training on our datasets. This highlights the challenge of finding a suitable tradeoff between achieving performance on specific tasks while preserving the original prowess of LLMs. Empirical findings suggest that tasks inherently equipped with reasoning paths contribute significantly to preserving certain capabilities of LLMs against potential declines. Motivated by this, we introduce the Reasoning-augmented Continual Learning (RCL) approach. RCL integrates task-specific cues with meta-rationales, effectively reducing catastrophic forgetting in LLMs while expediting convergence on novel tasks

    Investigation of scaling effect on power factor of permanent magnet Vernier machines for wind power application

    Get PDF
    This study investigates the scaling effect on power factor of surface mounted permanent magnet Vernier (SPM-V) machines with power ratings ranging from 3 kW, 500 kW, 3 MW to 10 MW. For each power rating, different slot/pole number combinations have been considered to study the influence of key parameters including inter-pole magnet leakage and stator slot leakage on power factor. A detailed analytical modelling, incorporating these key parameters, is presented and validated with two-dimensional finite-element analysis for different power ratings and slot/pole number combinations. The study has revealed that with scaling (increasing power level), significant increase in electrical loading combined with the increased leakage fluxes, i.e. (i) magnet leakage flux due to large coil pitch to rotor pole pitch ratio, (ii) magnet inter-pole leakage flux and (iii) stator slot leakage flux, reduces the ratio of armature flux linkage to permanent magnet flux linkage and thereby has a detrimental effect on the power factor. Therefore, unlike conventional SPM machines, the power factor of SPM-V machines is found to be significantly reduced at high power ratings
    corecore